Skip to main content

Self-hosted server for Agent Debugger — FastAPI backend, SQLite/Postgres storage, SSE streaming, and React UI.

Project description

Peaky Peek

Local-first agent debugger with replay, failure memory, smart highlights, and drift detection.

pip install peaky-peek-server && peaky-peek --open

Local-first, open-source agent debugger. Capture decisions, replay from checkpoints, visualize reasoning trees — all on your machine, no data sent anywhere.

PyPI PyPI Server Python 3.10+ License CI Downloads


Peaky Peek demo walkthrough

Why Peaky Peek?

Traditional observability tools weren't built for agent-native debugging:

Tool Focus Problem
LangSmith LLM tracing SaaS-first, your data leaves your machine
OpenTelemetry Infra metrics Blind to reasoning chains and decision trees
Sentry Error tracking No insight into why agents chose specific actions
Peaky Peek Agent-native debugging Local-first, open source, privacy by default

Peaky Peek captures the causal chain behind every action so you can debug agents like distributed systems: trace failures, replay from checkpoints, and search across reasoning paths.


Quick Start

Option 1: Decorator (simplest)

pip install peaky-peek-server
peaky-peek --open   # launches API + UI at http://localhost:8000
from agent_debugger_sdk import trace

@trace
async def my_agent(prompt: str) -> str:
    # Your agent logic here — traces are captured automatically
    return await llm_call(prompt)

Option 2: Context Manager

from agent_debugger_sdk import trace_session

async with trace_session("weather_agent") as ctx:
    await ctx.record_decision(
        reasoning="User asked for weather",
        confidence=0.9,
        chosen_action="call_weather_api",
        evidence=[{"source": "user_input", "content": "What's the weather?"}],
    )
    await ctx.record_tool_call("weather_api", {"city": "Seattle"})
    await ctx.record_tool_result("weather_api", result={"temp": 52, "forecast": "rain"})

Option 3: Zero-Config Auto-Patch (no code changes)

# Set env var, then run your agent normally
PEAKY_PEEK_AUTO_PATCH=true python my_agent.py

Works with PydanticAI, LangChain, OpenAI SDK, CrewAI, AutoGen, LlamaIndex, and Anthropic — no imports or decorators needed.


Framework Integrations

PydanticAI

from pydantic_ai import Agent
from agent_debugger_sdk import init
from agent_debugger_sdk.adapters import PydanticAIAdapter

init()

agent = Agent("openai:gpt-4o")
adapter = PydanticAIAdapter(agent, agent_name="support_agent")

LangChain

from agent_debugger_sdk import init
from agent_debugger_sdk.adapters import LangChainTracingHandler

init()

handler = LangChainTracingHandler(session_id="my-session")
# Pass handler to your LangChain agent's callbacks

OpenAI SDK

No code needed — just set the environment variable:

PEAKY_PEEK_AUTO_PATCH=true python my_openai_agent.py

Or use the simplified decorator:

from agent_debugger_sdk import trace

@trace(name="openai_agent", framework="openai")
async def my_agent(prompt: str) -> str:
    client = openai.AsyncOpenAI()
    response = await client.chat.completions.create(
        model="gpt-4o", messages=[{"role": "user", "content": prompt}]
    )
    return response.choices[0].message.content

Auto-Patch (Any Framework)

import agent_debugger_sdk.auto_patch  # activates on import when PEAKY_PEEK_AUTO_PATCH is set

# Now run your agent normally — all LLM calls are traced automatically

Features

Decision Tree Visualization

Decision Tree visualization demo

Navigate agent reasoning as an interactive tree. Click nodes to inspect events, zoom to explore complex flows, and trace the causal chain from policy to tool call to safety check.

Checkpoint Replay

Checkpoint replay demo

Time-travel through agent execution with checkpoint-aware playback. Play, pause, step, and seek to any point in the trace. Checkpoints are ranked by restore value so you jump to the most useful state.

Trace Search

Find specific events across all sessions. Search by keyword, filter by event type, and jump directly to results.

Failure Clustering & Multi-Agent Coordination

Failure clustering demo

Multi-agent coordination demo

Adaptive analysis groups similar failures. Inspect planner/critic debates, speaker topology, and prompt policy parameters across multi-agent systems.

Session Comparison

Compare two agent runs side-by-side. See diffs in turn count, speaker topology, policies, stance shifts, and grounded decisions.


Privacy & Security

  • Local-first by default — no external telemetry, no data leaves your machine
  • Zero-config auto-patching — no credentials or API keys needed for local debugging
  • Optional redaction pipeline — prompts, payloads, PII regex
  • API key authentication — bcrypt hashing
  • GDPR/HIPAA friendly — SQLite storage, no cloud dependency

Deployment

pip (recommended)

pip install peaky-peek-server
peaky-peek --open

Docker

docker build -t peaky-peek .
docker run -p 8000:8000 -v ./traces:/app/traces peaky-peek

Development

git clone https://github.com/acailic/agent_debugger
cd agent_debugger
pip install -e ".[dev]"
pip install fastapi "uvicorn[standard]" "sqlalchemy[asyncio]" aiosqlite alembic aiofiles bcrypt
python3 -m pytest -q
cd frontend && npm install && npm run build

Architecture

flowchart TB
    subgraph SDK["SDK Layer"]
        direction LR
        DEC["@trace decorator"]
        CTX["trace_session()"]
        AP["Auto-Patch"]
        AD["Framework Adapters"]
    end

    subgraph API["API Layer — FastAPI"]
        direction LR
        R1["Sessions"]
        R2["Traces"]
        R3["Replay"]
        R4["Search"]
        R5["Analytics"]
        SSE["SSE Stream"]
    end

    subgraph STORE["Storage Layer"]
        direction LR
        S1["Sessions"]
        S2["Events"]
        S3["Checkpoints"]
        S4["Snapshots"]
    end

    subgraph UI["Frontend — React + TypeScript"]
        direction LR
        DT["Decision Tree"]
        TI["Tool Inspector"]
        SR["Session Replay"]
        FC["Failure Clustering"]
        MA["Multi-Agent View"]
    end

    SDK -- "HTTP / WebSocket" --> API
    API -- "SQLite / PostgreSQL" --> STORE
    UI -- "REST + SSE" --> API

See ARCHITECTURE.md for full module breakdown.


Project Status

  • Core debugger — local path end-to-end, stable
  • SDK@trace, trace_session(), auto-patch for 7 frameworks
  • API — 11 routers: sessions, traces, replay, search, analytics, cost, comparison
  • Frontend — 8 specialized panels (decision tree, replay, checkpoints, search)
  • Tests — 365+ passing, CI on Python 3.10/3.11/3.12

Scientific Foundations

Peaky Peek is informed by research on agent debugging, causal tracing, failure analysis, and adaptive replay. See paper notes for design takeaways from each.

Documentation


Contributing

Contributions are welcome! See CONTRIBUTING.md for guidelines.


License

MIT

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

peaky_peek_server-0.1.15.tar.gz (36.6 MB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

peaky_peek_server-0.1.15-py3-none-any.whl (781.8 kB view details)

Uploaded Python 3

File details

Details for the file peaky_peek_server-0.1.15.tar.gz.

File metadata

  • Download URL: peaky_peek_server-0.1.15.tar.gz
  • Upload date:
  • Size: 36.6 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for peaky_peek_server-0.1.15.tar.gz
Algorithm Hash digest
SHA256 da54dff672af838e4e811dd14b7bff68b8ce5468613d5175204ecd783f110c6d
MD5 3defc824248ce410fb4042cdac3aa013
BLAKE2b-256 b33e69cc05f693b90f83d6473b9d8a42806e80b37cddd704320e98337d97c812

See more details on using hashes here.

Provenance

The following attestation bundles were made for peaky_peek_server-0.1.15.tar.gz:

Publisher: publish.yml on acailic/agent_debugger

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file peaky_peek_server-0.1.15-py3-none-any.whl.

File metadata

File hashes

Hashes for peaky_peek_server-0.1.15-py3-none-any.whl
Algorithm Hash digest
SHA256 af371b3cc314e8f8e8459a2a6f805a919b5ebdf721096f5ac687eeeaf3c0a4e5
MD5 2b967aaec3a6a5365ab9eaa81b84993b
BLAKE2b-256 3aef5d163c4bea19adf6c4e806bb0d65fd1ae1f558691c131a1ca58344c91411

See more details on using hashes here.

Provenance

The following attestation bundles were made for peaky_peek_server-0.1.15-py3-none-any.whl:

Publisher: publish.yml on acailic/agent_debugger

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page